Fast leave-one-out evaluation for dynamic gene selection
نویسندگان
چکیده
Gene selection procedure is a necessary step to increase the accuracy of machine learning algorithms that help in disease diagnosis based on gene expression data. This is commonly known as a feature subset selection problem in machine learning domain. A fast leave-one-out (LOO) evaluation formula for least-squares support vector machines (LSSVMs) is introduced here that can guide our backward feature selection process. Based on that, we propose a fast LOO guided feature selection (LGFS) algorithm. The gene selection step size is dynamically adjusted according to the LOO accuracy estimation. For our experiments, the application of LGFS to the gene selection process improves the classifier accuracy and reduces the number of features required as well. The least number of genes that can maximize the disease classification accuracy is automatically determined by our algorithm.
منابع مشابه
The ties problem resulting from counting-based error estimators and its impact on gene selection algorithms
MOTIVATION Feature selection approaches, such as filter and wrapper, have been applied to address the gene selection problem in the literature of microarray data analysis. In wrapper methods, the classification error is usually used as the evaluation criterion of feature subsets. Due to the nature of high dimensionality and small sample size of microarray data, however, counting-based error est...
متن کاملLEAF: Leave-one-out Forward Selection Method for Gene Selection in DNA Microarray Data
Preventing, diagnosing, and treating disease is greatly facilitated by the availability of biomarkers. Recent improvements in bioinformatics technology have facilitated large-scale screening of DNA arrays for candidate biomarkers. Here we discuss a gene analysis method that we call the LEAve-one-out Forward selection method (LEAF) for discovering informative genes embedded in expression data, a...
متن کاملEvaluating Evaluation Measures
Model selection measures such as hold-out set error, cross-validation error, leave-one-out error etc. are used to evaluate the performance of a classification algorithm on a given data set. To get an accurate estimate of the performance it is important that we choose the appropriate model selection measure or evaluation measure for the setting of interest. In this paper, we describe in brief a ...
متن کاملFeature selection using genetic algorithm for classification of schizophrenia using fMRI data
In this paper we propose a new method for classification of subjects into schizophrenia and control groups using functional magnetic resonance imaging (fMRI) data. In the preprocessing step, the number of fMRI time points is reduced using principal component analysis (PCA). Then, independent component analysis (ICA) is used for further data analysis. It estimates independent components (ICs) of...
متن کاملA fast linear-in-the-parameters classifier construction algorithm using orthogonal forward selection to minimize leave-one-out misclassification rate
International Journal of Systems Science Publication details, including instructions for authors and subscription information: http://www.informaworld.com/smpp/title~content=t713697751 A fast linear-in-the-parameters classifier construction algorithm using orthogonal forward selection to minimize leave-one-out misclassification rate X. Hong a; S. Chen b; C. J. Harris b a School of Systems Engin...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Soft Comput.
دوره 10 شماره
صفحات -
تاریخ انتشار 2006